183 research outputs found

    Paradigms for Parameterized Enumeration

    Full text link
    The aim of the paper is to examine the computational complexity and algorithmics of enumeration, the task to output all solutions of a given problem, from the point of view of parameterized complexity. First we define formally different notions of efficient enumeration in the context of parameterized complexity. Second we show how different algorithmic paradigms can be used in order to get parameter-efficient enumeration algorithms in a number of examples. These paradigms use well-known principles from the design of parameterized decision as well as enumeration techniques, like for instance kernelization and self-reducibility. The concept of kernelization, in particular, leads to a characterization of fixed-parameter tractable enumeration problems.Comment: Accepted for MFCS 2013; long version of the pape

    A shortcut to (sun)flowers: Kernels in logarithmic space or linear time

    Full text link
    We investigate whether kernelization results can be obtained if we restrict kernelization algorithms to run in logarithmic space. This restriction for kernelization is motivated by the question of what results are attainable for preprocessing via simple and/or local reduction rules. We find kernelizations for d-Hitting Set(k), d-Set Packing(k), Edge Dominating Set(k) and a number of hitting and packing problems in graphs, each running in logspace. Additionally, we return to the question of linear-time kernelization. For d-Hitting Set(k) a linear-time kernelization was given by van Bevern [Algorithmica (2014)]. We give a simpler procedure and save a large constant factor in the size bound. Furthermore, we show that we can obtain a linear-time kernel for d-Set Packing(k) as well.Comment: 18 page

    A structural approach to kernels for ILPs: Treewidth and Total Unimodularity

    Get PDF
    Kernelization is a theoretical formalization of efficient preprocessing for NP-hard problems. Empirically, preprocessing is highly successful in practice, for example in state-of-the-art ILP-solvers like CPLEX. Motivated by this, previous work studied the existence of kernelizations for ILP related problems, e.g., for testing feasibility of Ax <= b. In contrast to the observed success of CPLEX, however, the results were largely negative. Intuitively, practical instances have far more useful structure than the worst-case instances used to prove these lower bounds. In the present paper, we study the effect that subsystems with (Gaifman graph of) bounded treewidth or totally unimodularity have on the kernelizability of the ILP feasibility problem. We show that, on the positive side, if these subsystems have a small number of variables on which they interact with the remaining instance, then we can efficiently replace them by smaller subsystems of size polynomial in the domain without changing feasibility. Thus, if large parts of an instance consist of such subsystems, then this yields a substantial size reduction. We complement this by proving that relaxations to the considered structures, e.g., larger boundaries of the subsystems, allow worst-case lower bounds against kernelization. Thus, these relaxed structures can be used to build instance families that cannot be efficiently reduced, by any approach.Comment: Extended abstract in the Proceedings of the 23rd European Symposium on Algorithms (ESA 2015

    Constraint satisfaction parameterized by solution size

    Full text link
    In the constraint satisfaction problem (CSP) corresponding to a constraint language (i.e., a set of relations) Γ\Gamma, the goal is to find an assignment of values to variables so that a given set of constraints specified by relations from Γ\Gamma is satisfied. The complexity of this problem has received substantial amount of attention in the past decade. In this paper we study the fixed-parameter tractability of constraint satisfaction problems parameterized by the size of the solution in the following sense: one of the possible values, say 0, is "free," and the number of variables allowed to take other, "expensive," values is restricted. A size constraint requires that exactly kk variables take nonzero values. We also study a more refined version of this restriction: a global cardinality constraint prescribes how many variables have to be assigned each particular value. We study the parameterized complexity of these types of CSPs where the parameter is the required number kk of nonzero variables. As special cases, we can obtain natural and well-studied parameterized problems such as Independent Set, Vertex Cover, d-Hitting Set, Biclique, etc. In the case of constraint languages closed under substitution of constants, we give a complete characterization of the fixed-parameter tractable cases of CSPs with size constraints, and we show that all the remaining problems are W[1]-hard. For CSPs with cardinality constraints, we obtain a similar classification, but for some of the problems we are only able to show that they are Biclique-hard. The exact parameterized complexity of the Biclique problem is a notorious open problem, although it is believed to be W[1]-hard.Comment: To appear in SICOMP. Conference version in ICALP 201

    On the Computational Complexity of Vertex Integrity and Component Order Connectivity

    Full text link
    The Weighted Vertex Integrity (wVI) problem takes as input an nn-vertex graph GG, a weight function w:V(G)Nw:V(G)\to\mathbb{N}, and an integer pp. The task is to decide if there exists a set XV(G)X\subseteq V(G) such that the weight of XX plus the weight of a heaviest component of GXG-X is at most pp. Among other results, we prove that: (1) wVI is NP-complete on co-comparability graphs, even if each vertex has weight 11; (2) wVI can be solved in O(pp+1n)O(p^{p+1}n) time; (3) wVI admits a kernel with at most p3p^3 vertices. Result (1) refutes a conjecture by Ray and Deogun and answers an open question by Ray et al. It also complements a result by Kratsch et al., stating that the unweighted version of the problem can be solved in polynomial time on co-comparability graphs of bounded dimension, provided that an intersection model of the input graph is given as part of the input. An instance of the Weighted Component Order Connectivity (wCOC) problem consists of an nn-vertex graph GG, a weight function w:V(G)Nw:V(G)\to \mathbb{N}, and two integers kk and ll, and the task is to decide if there exists a set XV(G)X\subseteq V(G) such that the weight of XX is at most kk and the weight of a heaviest component of GXG-X is at most ll. In some sense, the wCOC problem can be seen as a refined version of the wVI problem. We prove, among other results, that: (4) wCOC can be solved in O(min{k,l}n3)O(\min\{k,l\}\cdot n^3) time on interval graphs, while the unweighted version can be solved in O(n2)O(n^2) time on this graph class; (5) wCOC is W[1]-hard on split graphs when parameterized by kk or by ll; (6) wCOC can be solved in 2O(klogl)n2^{O(k\log l)} n time; (7) wCOC admits a kernel with at most kl(k+l)+kkl(k+l)+k vertices. We also show that result (6) is essentially tight by proving that wCOC cannot be solved in 2o(klogl)nO(1)2^{o(k \log l)}n^{O(1)} time, unless the ETH fails.Comment: A preliminary version of this paper already appeared in the conference proceedings of ISAAC 201

    Reduction Techniques for Graph Isomorphism in the Context of Width Parameters

    Full text link
    We study the parameterized complexity of the graph isomorphism problem when parameterized by width parameters related to tree decompositions. We apply the following technique to obtain fixed-parameter tractability for such parameters. We first compute an isomorphism invariant set of potential bags for a decomposition and then apply a restricted version of the Weisfeiler-Lehman algorithm to solve isomorphism. With this we show fixed-parameter tractability for several parameters and provide a unified explanation for various isomorphism results concerned with parameters related to tree decompositions. As a possibly first step towards intractability results for parameterized graph isomorphism we develop an fpt Turing-reduction from strong tree width to the a priori unrelated parameter maximum degree.Comment: 23 pages, 4 figure

    On the stable degree of graphs

    No full text
    We define the stable degree s(G) of a graph G by s(G)∈=∈ min max d (v), where the minimum is taken over all maximal independent sets U of G. For this new parameter we prove the following. Deciding whether a graph has stable degree at most k is NP-complete for every fixed k∈≥∈3; and the stable degree is hard to approximate. For asteroidal triple-free graphs and graphs of bounded asteroidal number the stable degree can be computed in polynomial time. For graphs in these classes the treewidth is bounded from below and above in terms of the stable degree

    On Directed Feedback Vertex Set parameterized by treewidth

    Get PDF
    We study the Directed Feedback Vertex Set problem parameterized by the treewidth of the input graph. We prove that unless the Exponential Time Hypothesis fails, the problem cannot be solved in time 2o(tlogt)nO(1)2^{o(t\log t)}\cdot n^{\mathcal{O}(1)} on general directed graphs, where tt is the treewidth of the underlying undirected graph. This is matched by a dynamic programming algorithm with running time 2O(tlogt)nO(1)2^{\mathcal{O}(t\log t)}\cdot n^{\mathcal{O}(1)}. On the other hand, we show that if the input digraph is planar, then the running time can be improved to 2O(t)nO(1)2^{\mathcal{O}(t)}\cdot n^{\mathcal{O}(1)}.Comment: 20
    corecore